Global Convergence of Conjugate Gradient Methods without Line Search
نویسندگان
چکیده
Global convergence results are derived for well-known conjugate gradient methods in which the line search step is replaced by a step whose length is determined by a formula. The results include the following cases: 1. The Fletcher-Reeves method, the Hestenes-Stiefel method, and the Dai-Yuan method applied to a strongly convex LC objective function; 2. The Polak-Ribière method and the Conjugate Descent method applied to a general, not necessarily convex, LC objective function.
منابع مشابه
A New Hybrid Conjugate Gradient Method Based on Eigenvalue Analysis for Unconstrained Optimization Problems
In this paper, two extended three-term conjugate gradient methods based on the Liu-Storey ({tt LS}) conjugate gradient method are presented to solve unconstrained optimization problems. A remarkable property of the proposed methods is that the search direction always satisfies the sufficient descent condition independent of line search method, based on eigenvalue analysis. The globa...
متن کاملA new hybrid conjugate gradient algorithm for unconstrained optimization
In this paper, a new hybrid conjugate gradient algorithm is proposed for solving unconstrained optimization problems. This new method can generate sufficient descent directions unrelated to any line search. Moreover, the global convergence of the proposed method is proved under the Wolfe line search. Numerical experiments are also presented to show the efficiency of the proposed algorithm, espe...
متن کاملA Three-terms Conjugate Gradient Algorithm for Solving Large-Scale Systems of Nonlinear Equations
Nonlinear conjugate gradient method is well known in solving large-scale unconstrained optimization problems due to it’s low storage requirement and simple to implement. Research activities on it’s application to handle higher dimensional systems of nonlinear equations are just beginning. This paper presents a Threeterm Conjugate Gradient algorithm for solving Large-Scale systems of nonlinear e...
متن کاملExtensions of the Hestenes-Stiefel and Polak-Ribiere-Polyak conjugate gradient methods with sufficient descent property
Using search directions of a recent class of three--term conjugate gradient methods, modified versions of the Hestenes-Stiefel and Polak-Ribiere-Polyak methods are proposed which satisfy the sufficient descent condition. The methods are shown to be globally convergent when the line search fulfills the (strong) Wolfe conditions. Numerical experiments are done on a set of CUTEr unconstrained opti...
متن کاملOn Convergence and Restart Conditions for a Nonlinear Conjugate Gradient Method
A global convergence theorem for unconstrained minimization algorithms with an eecient line search is derived. The theorem applies to a new version of the conjugate gradient method derived here in terms of minimizing the eeect of zigzagging. The global convergence condition makes much weaker demands on the line search than previous methods. Local Q-linear convergence in a neighborhood of a stro...
متن کاملA new class of Conjugate Gradient Methods with extended Nonmonotone Line Search
In this paper, we propose a new nonlinear conjugate gradient method for large-scale unconstrain optimization which possesses the following properties:(i)the sufficient descent condition −g k dk ≥ 7 8 ‖gk‖ 2 holds without any line searches;(ii)With exact line search, this method reduces to a nonlinear version of the Liu-Storey conjugate gradient scheme.(iii)Under some assumption, global converge...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید
ثبت ناماگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید
ورودعنوان ژورنال:
- Annals OR
دوره 103 شماره
صفحات -
تاریخ انتشار 2001